[Amazon FSx for NetApp ONTAP] SnapMirrorの転送先ボリュームのサイズ不足である場合の挙動を確認してみた

[Amazon FSx for NetApp ONTAP] SnapMirrorの転送先ボリュームのサイズ不足である場合の挙動を確認してみた

Clock Icon2023.12.28

SnapMirrorの転送先ボリュームに空きがなかった場合の挙動が気になる

こんにちは、のんピ(@non____97)です。

皆さんはSnapMirrorの転送先ボリュームに空きがなかった場合の挙動が気になったことはありますか? 私はあります。

SnapMirrorを使ってAmazon FSx for NetApp ONTAP(以降FSxN)で移行する場合、コスト削減のためにSSDの使用量を抑えたいと考えるかと思います。

そのような場合に以下のような方法が使えないかと考えました。

  1. SSDのサイズを小さめに作成する
  2. SnapMirrorでできる限り転送する
  3. 途中で中断されたらStorage Efficiencyを実行して空きを作成する
  4. 2と3の処理を繰り返し実行する

「キャパシティプールストレージに流せば良いのでは?」と思われるかもしれませんが、重複排除や圧縮などのStorage EfficiencyはSSD上でしかかけることができません。

実際に上述のようなことが可能なのか確認してみました。

いきなりまとめ

  • 冒頭で紹介した方式は実現できない
  • SnapMirrorの転送先ボリュームのサイズ不足である場合、転送が中止される
  • SnapMirrorのInitializeで途中で終了した場合、転送先ボリュームでStorage Efficiencyを有効化することはできない
    • ボリュームサイズを拡張するなど問題を解決した上で再度Initializeした後に操作する必要がある
    • つまり、転送先ボリュームに十分な空き容量がなければ、Storage Efficiencyを実行することができない
  • SnapMirrorの差分同期途中で転送が終了した場合、転送先ボリュームでStorage Efficiencyを有効化することはできない
    • ボリュームサイズを拡張するなど問題を解決した上でチェックポイントを削除した後に操作する必要がある
  • SnapMirrorの転送先ボリュームのサイズは転送元ボリュームのサイズ以上のものになるようにしよう
    • ボリュームの自動拡張の有効化がオススメ
  • 少ないSSDで重複排除や圧縮などのStorage Efficiencyをしっかりと効かせたい場合は、カスケードSnapMirrorを使う方式となる

やってみる

検証環境の確認

検証環境は以下記事で用意したものを利用します。

改めて転送元ボリュームやStorage Efficiency、aggregate、Snapshotを確認しておきます。

::> set diag

Warning: These diagnostic commands are for use by NetApp personnel only.
Do you want to continue? {y|n}: y

::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression
vserver volume state   policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume
------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- -----------------------------------------
svm     vol1   Enabled auto   false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst
               Enabled -      false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst2
               Enabled -      false       true               efficient               false         true            true                              true                          false
svm3    vol1_dst2_dst
               Enabled auto   false       true               efficient               true          true            true                              true                          false
svm3    vol1_dst_dst
               Enabled auto   false       true               efficient               true          true            true                              true                          false
5 entries were displayed.

::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 6.85GB    16GB            15.20GB 8.35GB 54%          808.8MB            9%                         808.8MB             9.14GB       60%                  -                 9.14GB              -                                   -
svm2    vol1_dst
               7.54GB
                    1.10GB    7.54GB          7.16GB  6.06GB 84%          4.98GB             45%                        3.00GB              11.01GB      154%                 -                 8.05GB              0B                                  0%
svm2    vol1_dst2
               9.84GB
                    1.45GB    9.84GB          9.35GB  7.90GB 84%          4.53GB             36%                        4.10GB              12.40GB      133%                 -                 9.06GB              0B                                  0%
svm3    vol1_dst2_dst
               5.40GB
                    876.9MB   5.40GB          5.40GB  4.55GB 84%          4.59GB             50%                        2.05GB              9.13GB       169%                 -                 9.04GB              0B                                  0%
svm3    vol1_dst_dst
               4.90GB
                    896.3MB   4.90GB          4.90GB  4.03GB 82%          5.11GB             56%                        2.21GB              9.14GB       186%                 -                 9.07GB              0B                                  0%
5 entries were displayed.

::*> volume show-footprint -volume vol1


      Vserver : svm
      Volume  : vol1

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            8.38GB       1%
             Footprint in Performance Tier            245.6MB       3%
             Footprint in FSxFabricpoolObjectStore
                                                       8.24GB      97%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                        92.66MB       0%
      Deduplication Metadata                          12.04MB       0%
           Deduplication                              12.04MB       0%
      Delayed Frees                                   111.0MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  8.58GB       1%

      Footprint Data Reduction in capacity tier        3.87GB        -
      Effective Total Footprint                        4.71GB       1%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 196.4GB
                               Total Physical Used: 19.87GB
                    Total Storage Efficiency Ratio: 9.88:1
Total Data Reduction Logical Used Without Snapshots: 43.87GB
Total Data Reduction Physical Used Without Snapshots: 14.80GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.96:1
Total Data Reduction Logical Used without snapshots and flexclones: 43.87GB
Total Data Reduction Physical Used without snapshots and flexclones: 14.80GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.96:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 141.9GB
Total Physical Used in FabricPool Performance Tier: 16.29GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 8.71:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.57GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 11.23GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.17:1
                Logical Space Used for All Volumes: 43.87GB
               Physical Space Used for All Volumes: 23.87GB
               Space Saved by Volume Deduplication: 20.00GB
Space Saved by Volume Deduplication and pattern detection: 20.00GB
                Volume Deduplication Savings ratio: 1.84:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.84:1
               Logical Space Used by the Aggregate: 28.82GB
              Physical Space Used by the Aggregate: 19.87GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.45:1
              Logical Size Used by Snapshot Copies: 152.6GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 20.73:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 20.73:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 0

::*> snapshot show -volume vol1
                                                                 ---Blocks---
Vserver  Volume   Snapshot                                  Size Total% Used%
-------- -------- ------------------------------------- -------- ------ -----
svm      vol1
                  test.2023-12-22_0533                     160KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                         24.45MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           312KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           144KB     0%    0%
6 entries were displayed.

転送先ボリュームの作成

転送先ボリュームを作成します。

転送元ボリュームの使用量は8.35GBです。転送先ボリュームのサイズは2GBで自動拡張しないようにします。

volume create -vserver svm2 -volume vol1_dst3 -aggregate aggr1 -state online -type DP -size 2GB -tiering-policy none 

::*> volume create -vserver svm2 -volume vol1_dst3 -aggregate aggr1 -state online -type DP -size 2GB -tiering-policy none -autosize-mode off
[Job 149] Job succeeded: Successful

::*> volume show -volume vol1* -fields type, autosize-mode, max-autosize
vserver volume max-autosize autosize-mode type
------- ------ ------------ ------------- ----
svm     vol1   19.20GB      off           RW
svm2    vol1_dst
               100TB        grow_shrink   DP
svm2    vol1_dst2
               100TB        grow_shrink   DP
svm2    vol1_dst3
               100TB        off           DP
svm3    vol1_dst2_dst
               100TB        grow_shrink   RW
svm3    vol1_dst_dst
               100TB        grow_shrink   RW
6 entries were displayed.

SnapMirror Initialize

SnapMirrorのInitializeを行います。

下準備としてSnapMirror relationshipを作成します。

::*> snapmirror create -source-path svm:vol1 -destination-path svm2:vol1_dst3 -policy MirrorAllSnapshots
Operation succeeded: snapmirror create for the relationship with destination "svm2:vol1_dst3".

::*> snapmirror show
                                                                       Progress
Source            Destination Mirror  Relationship   Total             Last
Path        Type  Path        State   Status         Progress  Healthy Updated
----------- ---- ------------ ------- -------------- --------- ------- --------
svm:vol1    XDP  svm2:vol1_dst
                              Snapmirrored
                                      Idle           -         true    -
                 svm2:vol1_dst2
                              Snapmirrored
                                      Idle           -         true    -
                 svm2:vol1_dst3
                              Uninitialized
                                      Idle           -         true    -
                 svm3:vol1_dst_dst
                              Broken-off
                                      Idle           -         true    -
svm2:vol1_dst2
            XDP  svm3:vol1_dst2_dst
                              Broken-off
                                      Idle           -         true    -
5 entries were displayed.

::*> volume show -volume vol1, vol1_dst3 -fields type, autosize-mode, max-autosize
vserver volume max-autosize autosize-mode type
------- ------ ------------ ------------- ----
svm     vol1   19.20GB      off           RW
svm2    vol1_dst3
               100TB        off           DP
2 entries were displayed.

Initializeします。

::*> snapmirror initialize -destination-path svm2:vol1_dst3
Operation is queued: snapmirror initialize of destination "svm2:vol1_dst3".

::*> snapmirror show -destination-path svm2:vol1_dst3

                                  Source Path: svm:vol1
                               Source Cluster: -
                               Source Vserver: svm
                                Source Volume: vol1
                             Destination Path: svm2:vol1_dst3
                          Destination Cluster: -
                          Destination Vserver: svm2
                           Destination Volume: vol1_dst3
                            Relationship Type: XDP
                      Relationship Group Type: none
                             Managing Vserver: svm2
                          SnapMirror Schedule: -
                       SnapMirror Policy Type: async-mirror
                            SnapMirror Policy: MirrorAllSnapshots
                                  Tries Limit: -
                            Throttle (KB/sec): unlimited
              Consistency Group Item Mappings: -
           Current Transfer Throttle (KB/sec): unlimited
                                 Mirror State: Uninitialized
                          Relationship Status: Transferring
                      File Restore File Count: -
                       File Restore File List: -
                            Transfer Snapshot: test.2023-12-22_0533
                            Snapshot Progress: 351.6MB
                               Total Progress: 351.6MB
                    Network Compression Ratio: 1:1
                          Snapshot Checkpoint: 594.6KB
                              Newest Snapshot: -
                    Newest Snapshot Timestamp: -
                            Exported Snapshot: -
                  Exported Snapshot Timestamp: -
                                      Healthy: true
                              Relationship ID: e94d0e31-a526-11ee-981e-bdd56ead09c8
                          Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8
                     Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8
                         Current Operation ID: 1b60dcad-a527-11ee-981e-bdd56ead09c8
                                Transfer Type: initialize
                               Transfer Error: -
                           Last Transfer Type: -
                          Last Transfer Error: -
                    Last Transfer Error Codes: -
                           Last Transfer Size: -
      Last Transfer Network Compression Ratio: -
                       Last Transfer Duration: -
                           Last Transfer From: -
                  Last Transfer End Timestamp: -
                             Unhealthy Reason: -
                        Progress Last Updated: 12/28 02:16:31
                      Relationship Capability: 8.2 and above
                                     Lag Time: -
                    Current Transfer Priority: normal
                             SMTape Operation: -
                 Destination Volume Node Name: FsxId0ab6f9b00824a187c-01
                 Identity Preserve Vserver DR: -
                 Number of Successful Updates: 0
                     Number of Failed Updates: 0
                 Number of Successful Resyncs: 0
                     Number of Failed Resyncs: 0
                  Number of Successful Breaks: 0
                      Number of Failed Breaks: 0
                         Total Transfer Bytes: 0
               Total Transfer Time in Seconds: 0
                Source Volume MSIDs Preserved: -
                                       OpMask: ffffffffffffffff
                       Is Auto Expand Enabled: -
          Percent Complete for Current Status: -

::*> snapmirror show
                                                                       Progress
Source            Destination Mirror  Relationship   Total             Last
Path        Type  Path        State   Status         Progress  Healthy Updated
----------- ---- ------------ ------- -------------- --------- ------- --------
svm:vol1    XDP  svm2:vol1_dst
                              Snapmirrored
                                      Idle           -         true    -
                 svm2:vol1_dst2
                              Snapmirrored
                                      Idle           -         true    -
                 svm2:vol1_dst3
                              Uninitialized
                                      Idle           -         false   -
                 svm3:vol1_dst_dst
                              Broken-off
                                      Idle           -         true    -
svm2:vol1_dst2
            XDP  svm3:vol1_dst2_dst
                              Broken-off
                                      Idle           -         true    -
5 entries were displayed.

::*> snapmirror show -destination-path svm2:vol1_dst3

                                  Source Path: svm:vol1
                               Source Cluster: -
                               Source Vserver: svm
                                Source Volume: vol1
                             Destination Path: svm2:vol1_dst3
                          Destination Cluster: -
                          Destination Vserver: svm2
                           Destination Volume: vol1_dst3
                            Relationship Type: XDP
                      Relationship Group Type: none
                             Managing Vserver: svm2
                          SnapMirror Schedule: -
                       SnapMirror Policy Type: async-mirror
                            SnapMirror Policy: MirrorAllSnapshots
                                  Tries Limit: -
                            Throttle (KB/sec): unlimited
              Consistency Group Item Mappings: -
           Current Transfer Throttle (KB/sec): -
                                 Mirror State: Uninitialized
                          Relationship Status: Idle
                      File Restore File Count: -
                       File Restore File List: -
                            Transfer Snapshot: -
                            Snapshot Progress: -
                               Total Progress: -
                    Network Compression Ratio: -
                          Snapshot Checkpoint: 594.6KB
                              Newest Snapshot: -
                    Newest Snapshot Timestamp: -
                            Exported Snapshot: -
                  Exported Snapshot Timestamp: -
                                      Healthy: false
                              Relationship ID: e94d0e31-a526-11ee-981e-bdd56ead09c8
                          Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8
                     Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8
                         Current Operation ID: -
                                Transfer Type: -
                               Transfer Error: -
                           Last Transfer Type: initialize
                          Last Transfer Error: Transfer failed. (Volume access error (No space left on device))
                    Last Transfer Error Codes: 6620144, 5898547, 6684700
                           Last Transfer Size: -
      Last Transfer Network Compression Ratio: -
                       Last Transfer Duration: -
                           Last Transfer From: svm:vol1
                  Last Transfer End Timestamp: 12/28 02:24:11
                             Unhealthy Reason: Transfer failed.
                        Progress Last Updated: -
                      Relationship Capability: 8.2 and above
                                     Lag Time: -
                    Current Transfer Priority: -
                             SMTape Operation: -
                 Destination Volume Node Name: FsxId0ab6f9b00824a187c-01
                 Identity Preserve Vserver DR: -
                 Number of Successful Updates: 0
                     Number of Failed Updates: 0
                 Number of Successful Resyncs: 0
                     Number of Failed Resyncs: 0
                  Number of Successful Breaks: 0
                      Number of Failed Breaks: 0
                         Total Transfer Bytes: 0
               Total Transfer Time in Seconds: 0
                Source Volume MSIDs Preserved: -
                                       OpMask: ffffffffffffffff
                       Is Auto Expand Enabled: -
          Percent Complete for Current Status: -

しばらく待つと、ボリュームの空き容量不足により転送が失敗していました。

EMSにもSnapMirrorの転送失敗は記録されていました。

::*> event log show
Time                Node             Severity      Event
------------------- ---------------- ------------- ---------------------------
12/28/2023 02:22:12 FsxId0ab6f9b00824a187c-01
                                     ERROR         monitor.volume.nearlyFull: Volume vol1_dst3@vserver:5af907bb-a065-11ee-981e-bdd56ead09c8 is nearly full(using or reserving 97% of space and 0% of inodes).
12/28/2023 02:20:09 FsxId0ab6f9b00824a187c-01
                                     ERROR         monitor.volume.nearlyFull: Volume vol1_dst3@vserver:5af907bb-a065-11ee-981e-bdd56ead09c8 is nearly full(using or reserving 97% of space and 0% of inodes).
12/28/2023 02:17:52 FsxId0ab6f9b00824a187c-01
                                     ERROR         monitor.volume.nearlyFull: Volume vol1_dst3@vserver:5af907bb-a065-11ee-981e-bdd56ead09c8 is nearly full(using or reserving 97% of space and 0% of inodes).
12/28/2023 02:12:44 FsxId0ab6f9b00824a187c-01
                                     NOTICE        arw.volume.state: Anti-ransomware state was changed to "disabled" on volume "vol1_dst3" (UUID: "95d186d3-a526-11ee-981e-bdd56ead09c8") in Vserver "svm2" (UUID: "5af907bb-a065-11ee-981e-bdd56ead09c8").
4 entries were displayed.

Storage Efficiency、ボリューム、aggregate、Snapshotを確認します。

::*> volume efficiency show -volume vol1, vol1_dst3 -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression
vserver volume state   policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume
------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- -----------------------------------------
svm     vol1   Enabled auto   false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst3
               Disabled
                       -      false       true               efficient               false         true            true                              true                          false
2 entries were displayed.

::*> volume show -volume vol1, vol1_dst3 -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 6.85GB    16GB            15.20GB 8.35GB 54%          808.8MB            9%                         808.8MB             9.14GB       60%                  -                 9.14GB              -                                   -
svm2    vol1_dst3
               2GB  40.79MB   2GB             2GB     1.96GB 98%          0B                 0%                         0B                  1.96GB       98%                  -                 1.96GB              0B                                  0%
2 entries were displayed.

::*> volume show-footprint -volume vol1_dst3


      Vserver : svm2
      Volume  : vol1_dst3

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            1.96GB       0%
             Footprint in Performance Tier             1.98GB     100%
             Footprint in FSxFabricpoolObjectStore         0B       0%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                        20.98MB       0%
      Delayed Frees                                   16.74MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  2.00GB       0%

      Effective Total Footprint                        2.00GB       0%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 207.5GB
                               Total Physical Used: 21.83GB
                    Total Storage Efficiency Ratio: 9.50:1
Total Data Reduction Logical Used Without Snapshots: 45.82GB
Total Data Reduction Physical Used Without Snapshots: 16.61GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.76:1
Total Data Reduction Logical Used without snapshots and flexclones: 45.82GB
Total Data Reduction Physical Used without snapshots and flexclones: 16.61GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.76:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 144.1GB
Total Physical Used in FabricPool Performance Tier: 18.27GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 7.89:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 37.53GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 13.06GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.87:1
                Logical Space Used for All Volumes: 45.82GB
               Physical Space Used for All Volumes: 25.82GB
               Space Saved by Volume Deduplication: 20.00GB
Space Saved by Volume Deduplication and pattern detection: 20.00GB
                Volume Deduplication Savings ratio: 1.77:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.77:1
               Logical Space Used by the Aggregate: 30.78GB
              Physical Space Used by the Aggregate: 21.83GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.41:1
              Logical Size Used by Snapshot Copies: 161.7GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 21.97:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 21.97:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 1

::*> snapshot show -volume vol1, vol1_dst3
                                                                 ---Blocks---
Vserver  Volume   Snapshot                                  Size Total% Used%
-------- -------- ------------------------------------- -------- ------ -----
svm      vol1
                  test.2023-12-22_0533                     160KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                         24.45MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           312KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           136KB     0%    0%
7 entries were displayed.

1.96GBまで書き込んでいますね。また、転送先ボリュームでSnapshotは作成されていません。

転送先ボリュームをNFSクライアントでマウント

転送先ボリュームをNFSクライアントでマウントして、様子を確認してみましょう。

転送先ボリュームをSVMのジャンクションパスにマウントします。

::*> volume mount -vserver svm2 -volume vol1_dst3 -junction-path /vol1_dst3
Queued private job: 66

Error: command failed: Volume vol1_dst3 in Vserver "svm2" is not mountable until a "snapmirror initialize" has been completed.

はい、SnapMirrorのInitializeが完了していないボリュームはマウントできないようです。

転送先ボリュームでStorage Efficiencyを実行

転送先ボリュームでStorage Efficiencyを実行して、空きが作成されるか確認しましょう。

まず、Storage Efficiencyを有効化します。

::*> volume efficiency on -vserver svm2 -volume vol1_dst
    vol1_dst  vol1_dst2 vol1_dst3
::*> volume efficiency on -vserver svm2 -volume vol1_dst3

Error: command failed: Failed to enable efficiency on volume "vol1_dst3" of Vserver "svm2": A SnapMirror transfer is running or paused. Use the
       "snapmirror show" command to view the status of the transfer. Retry this command when the transfer is complete, or run "snapmirror abort -hard
       true" to abort the SnapMirror transfer and clear any checkpoint data.

SnapMirrorの転送中や転送が中断されている場合はStorage Efficiencyを有効にできないようです。

指示に従ってsnapMirror abortを叩いてみます。

::*> snapmirror abort -destination-path svm2:vol1_dst3

Error: command failed: No transfer to abort.

::*> snapmirror abort -destination-path svm2:vol1_dst3 -hard true

Error: command failed: Deleting checkpoint for uninitialized relationship is not supported.

どちらも失敗しました。SnapMirrorのチェックポイントを削除しようにもSnapMirror Initializeが完了していないためできないようです。

ということで、冒頭の仮説では対応できないことが分かりました。

少ないSSDで重複排除や圧縮などのStorage Efficiencyをしっかり効かせたい場合は、以下記事のとおりカスケードSnapMirrorを使うと良いでしょう。

SnapMirror update時にボリュームの空き容量が足りない場合

ボリュームの拡張

SnapMirror Initializeの時は上手くいきませんでしたが、SnapMirror update時にボリュームの空き容量が足りない場合はどうでしょうか。

まず、ボリュームを拡張します。

::*> volume modify -vserver svm2 -volume vol1_dst3 -size 9GB
Volume modify successful on volume vol1_dst3 of Vserver svm2.

::*> snapmirror show
                                                                       Progress
Source            Destination Mirror  Relationship   Total             Last
Path        Type  Path        State   Status         Progress  Healthy Updated
----------- ---- ------------ ------- -------------- --------- ------- --------
svm:vol1    XDP  svm2:vol1_dst
                              Snapmirrored
                                      Idle           -         true    -
                 svm2:vol1_dst2
                              Snapmirrored
                                      Idle           -         true    -
                 svm2:vol1_dst3
                              Uninitialized
                                      Idle           -         false   -
                 svm3:vol1_dst_dst
                              Broken-off
                                      Idle           -         true    -
svm2:vol1_dst2
            XDP  svm3:vol1_dst2_dst
                              Broken-off
                                      Idle           -         true    -
5 entries were displayed.

SnapMirror Initialize

SnapMirrorのInitializeをします。

::*> snapmirror initialize -destination-path svm2:vol1_dst3Operation is queued: snapmirror initialize of destination "svm2:vol1_dst3".

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status total-progress progress-last-updated
----------- ---------------- ------------ ------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Idle   -              -


::*> snapmirror show -destination-path svm2:vol1_dst3

                                  Source Path: svm:vol1
                               Source Cluster: -
                               Source Vserver: svm
                                Source Volume: vol1
                             Destination Path: svm2:vol1_dst3
                          Destination Cluster: -
                          Destination Vserver: svm2
                           Destination Volume: vol1_dst3
                            Relationship Type: XDP
                      Relationship Group Type: none
                             Managing Vserver: svm2
                          SnapMirror Schedule: -
                       SnapMirror Policy Type: async-mirror
                            SnapMirror Policy: MirrorAllSnapshots
                                  Tries Limit: -
                            Throttle (KB/sec): unlimited
              Consistency Group Item Mappings: -
           Current Transfer Throttle (KB/sec): -
                                 Mirror State: Snapmirrored
                          Relationship Status: Idle
                      File Restore File Count: -
                       File Restore File List: -
                            Transfer Snapshot: -
                            Snapshot Progress: -
                               Total Progress: -
                    Network Compression Ratio: -
                          Snapshot Checkpoint: -
                              Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                    Newest Snapshot Timestamp: 12/28 02:16:28
                            Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                  Exported Snapshot Timestamp: 12/28 02:16:28
                                      Healthy: true
                              Relationship ID: e94d0e31-a526-11ee-981e-bdd56ead09c8
                          Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8
                     Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8
                         Current Operation ID: -
                                Transfer Type: -
                               Transfer Error: -
                           Last Transfer Type: update
                          Last Transfer Error: -
                    Last Transfer Error Codes: -
                           Last Transfer Size: 3.07GB
      Last Transfer Network Compression Ratio: 1:1
                       Last Transfer Duration: 0:1:1
                           Last Transfer From: svm:vol1
                  Last Transfer End Timestamp: 12/28 03:04:32
                             Unhealthy Reason: -
                        Progress Last Updated: -
                      Relationship Capability: 8.2 and above
                                     Lag Time: 0:49:31
                    Current Transfer Priority: -
                             SMTape Operation: -
                 Destination Volume Node Name: FsxId0ab6f9b00824a187c-01
                 Identity Preserve Vserver DR: -
                 Number of Successful Updates: 1
                     Number of Failed Updates: 0
                 Number of Successful Resyncs: 0
                     Number of Failed Resyncs: 0
                  Number of Successful Breaks: 0
                      Number of Failed Breaks: 0
                         Total Transfer Bytes: 9060359240
               Total Transfer Time in Seconds: 112
                Source Volume MSIDs Preserved: -
                                       OpMask: ffffffffffffffff
                       Is Auto Expand Enabled: -
          Percent Complete for Current Status: -

転送が完了しました。

Storage Efficiency、ボリューム、aggregate、Snapshotを確認します。

::*> volume efficiency show -volume vol1, vol1_dst3 -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression
vserver volume state   policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume
------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- -----------------------------------------
svm     vol1   Enabled auto   false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst3
               Disabled
                       -      false       true               efficient               false         true            true                              true                          false
2 entries were displayed.

::*> volume show -volume vol1, vol1_dst3 -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 6.85GB    16GB            15.20GB 8.35GB 54%          808.8MB            9%                         808.8MB             9.14GB       60%                  -                 9.14GB              -                                   -
svm2    vol1_dst3
               9GB  693.7MB   9GB             9GB     8.32GB 92%          774.5MB            8%                         8.24GB              9.05GB       101%                 -                 9.04GB              0B                                  0%
2 entries were displayed.

::*> volume show-footprint -volume vol1_dst3


      Vserver : svm2
      Volume  : vol1_dst3

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            8.32GB       1%
             Footprint in Performance Tier             8.41GB     100%
             Footprint in FSxFabricpoolObjectStore         0B       0%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                        56.82MB       0%
      Delayed Frees                                   91.26MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  8.47GB       1%

      Effective Total Footprint                        8.47GB       1%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 270.9GB
                               Total Physical Used: 28.19GB
                    Total Storage Efficiency Ratio: 9.61:1
Total Data Reduction Logical Used Without Snapshots: 52.93GB
Total Data Reduction Physical Used Without Snapshots: 22.60GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.34:1
Total Data Reduction Logical Used without snapshots and flexclones: 52.93GB
Total Data Reduction Physical Used without snapshots and flexclones: 22.60GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.34:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 207.5GB
Total Physical Used in FabricPool Performance Tier: 24.65GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 8.42:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 44.65GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 19.08GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.34:1
                Logical Space Used for All Volumes: 52.93GB
               Physical Space Used for All Volumes: 32.18GB
               Space Saved by Volume Deduplication: 20.76GB
Space Saved by Volume Deduplication and pattern detection: 20.76GB
                Volume Deduplication Savings ratio: 1.65:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.65:1
               Logical Space Used by the Aggregate: 37.13GB
              Physical Space Used by the Aggregate: 28.19GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.32:1
              Logical Size Used by Snapshot Copies: 218.0GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 29.60:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 29.60:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 1

::*> snapshot show -volume vol1, vol1_dst3
                                                                 ---Blocks---
Vserver  Volume   Snapshot                                  Size Total% Used%
-------- -------- ------------------------------------- -------- ------ -----
svm      vol1
                  test.2023-12-22_0533                     160KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                         24.45MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           312KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           136KB     0%    0%
svm2     vol1_dst3
                  test.2023-12-22_0533                    1.29MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                           384KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           388KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           284KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           232KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           220KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           144KB     0%    0%
14 entries were displayed.

8.32GB書き込まれていますね。

テスト用ファイルの追加

転送元ボリュームにテスト用ファイルを追加します。既にボリューム上に存在しているファイルをコピーします。

$ sudo mount -t nfs svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 /mnt/fsxn/vol1

$ ls -l /mnt/fsxn/vol1
total 9474408
-rw-r--r--. 1 root root 1073741824 Dec 22 01:46 1_padding_file
-rw-r--r--. 1 root root 1073741824 Dec 22 07:53 ABCDE_padding_file
-rw-r--r--. 1 root root 1073741824 Dec 22 05:28 a_padding_file
-rw-r--r--. 1 root root 1073741824 Dec 22 06:55 abcde_padding_file
-rw-r--r--. 1 root root 1073741824 Dec 22 01:47 urandom_block_file
-rw-r--r--. 1 root root 1073741824 Dec 22 05:02 urandom_block_file2
-rw-r--r--. 1 root root 1073741824 Dec 22 05:02 urandom_block_file2_copy
-rw-r--r--. 1 root root 1073741824 Dec 22 01:47 urandom_block_file_copy
-rw-r--r--. 1 root root 1073741824 Dec 22 06:41 urandom_block_file_copy2

$ sudo cp /mnt/fsxn/vol1/urandom_block_file /mnt/fsxn/vol1/urandom_block_file_copy3

ボリューム、aggregateを確認します。

::*> volume show -volume vol1, vol1_dst3 -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 5.83GB    16GB            15.20GB 9.37GB 61%          808.8MB            8%                         808.8MB             10.16GB      67%                  -                 10.16GB             -                                   -
svm2    vol1_dst3
               9GB  693.7MB   9GB             9GB     8.32GB 92%          774.5MB            8%                         8.24GB              9.05GB       101%                 -                 9.04GB              0B                                  0%
2 entries were displayed.

::*> volume show-footprint -volume vol1_dst3


      Vserver : svm2
      Volume  : vol1_dst3

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            8.32GB       1%
             Footprint in Performance Tier             8.41GB     100%
             Footprint in FSxFabricpoolObjectStore         0B       0%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                        56.82MB       0%
      Delayed Frees                                   91.40MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  8.47GB       1%

      Effective Total Footprint                        8.47GB       1%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 271.9GB
                               Total Physical Used: 29.34GB
                    Total Storage Efficiency Ratio: 9.27:1
Total Data Reduction Logical Used Without Snapshots: 53.94GB
Total Data Reduction Physical Used Without Snapshots: 23.70GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.28:1
Total Data Reduction Logical Used without snapshots and flexclones: 53.94GB
Total Data Reduction Physical Used without snapshots and flexclones: 23.70GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.28:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 207.5GB
Total Physical Used in FabricPool Performance Tier: 24.81GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 8.36:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 44.67GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 19.19GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.33:1
                Logical Space Used for All Volumes: 53.94GB
               Physical Space Used for All Volumes: 33.18GB
               Space Saved by Volume Deduplication: 20.76GB
Space Saved by Volume Deduplication and pattern detection: 20.76GB
                Volume Deduplication Savings ratio: 1.63:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.63:1
               Logical Space Used by the Aggregate: 38.29GB
              Physical Space Used by the Aggregate: 29.34GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.30:1
              Logical Size Used by Snapshot Copies: 218.0GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 29.60:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 29.60:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 1

SnapMirror Update

SnapMirrorの差分同期を行います。

::*> snapmirror update -destination-path svm2:vol1_dst3
Operation is queued: snapmirror update of destination "svm2:vol1_dst3".

::*>  snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status       total-progress progress-last-updated
----------- ---------------- ------------ ------------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Transferring 18.73MB        12/28 04:17:21

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status       total-progress progress-last-updated
----------- ---------------- ------------ ------------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Transferring 343.0MB        12/28 04:17:37

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status       total-progress progress-last-updated
----------- ---------------- ------------ ------------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Transferring 651.2MB        12/28 04:18:08

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status total-progress progress-last-updated
----------- ---------------- ------------ ------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Idle   -              -

::*> snapmirror show -destination-path svm2:vol1_dst3

                                  Source Path: svm:vol1
                               Source Cluster: -
                               Source Vserver: svm
                                Source Volume: vol1
                             Destination Path: svm2:vol1_dst3
                          Destination Cluster: -
                          Destination Vserver: svm2
                           Destination Volume: vol1_dst3
                            Relationship Type: XDP
                      Relationship Group Type: none
                             Managing Vserver: svm2
                          SnapMirror Schedule: -
                       SnapMirror Policy Type: async-mirror
                            SnapMirror Policy: MirrorAllSnapshots
                                  Tries Limit: -
                            Throttle (KB/sec): unlimited
              Consistency Group Item Mappings: -
           Current Transfer Throttle (KB/sec): -
                                 Mirror State: Snapmirrored
                          Relationship Status: Idle
                      File Restore File Count: -
                       File Restore File List: -
                            Transfer Snapshot: -
                            Snapshot Progress: -
                               Total Progress: -
                    Network Compression Ratio: -
                          Snapshot Checkpoint: 265.8KB
                              Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                    Newest Snapshot Timestamp: 12/28 02:16:28
                            Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                  Exported Snapshot Timestamp: 12/28 02:16:28
                                      Healthy: false
                              Relationship ID: e94d0e31-a526-11ee-981e-bdd56ead09c8
                          Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8
                     Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8
                         Current Operation ID: -
                                Transfer Type: -
                               Transfer Error: -
                           Last Transfer Type: update
                          Last Transfer Error: Transfer failed. (Volume access error (No space left on device))
                    Last Transfer Error Codes: 6620144, 5898547, 6684700
                           Last Transfer Size: -
      Last Transfer Network Compression Ratio: -
                       Last Transfer Duration: -
                           Last Transfer From: svm:vol1
                  Last Transfer End Timestamp: 12/28 04:26:04
                             Unhealthy Reason: Transfer failed.
                        Progress Last Updated: -
                      Relationship Capability: 8.2 and above
                                     Lag Time: 2:34:18
                    Current Transfer Priority: -
                             SMTape Operation: -
                 Destination Volume Node Name: FsxId0ab6f9b00824a187c-01
                 Identity Preserve Vserver DR: -
                 Number of Successful Updates: 1
                     Number of Failed Updates: 1
                 Number of Successful Resyncs: 0
                     Number of Failed Resyncs: 0
                  Number of Successful Breaks: 0
                      Number of Failed Breaks: 0
                         Total Transfer Bytes: 9060359240
               Total Transfer Time in Seconds: 112
                Source Volume MSIDs Preserved: -
                                       OpMask: ffffffffffffffff
                       Is Auto Expand Enabled: -
          Percent Complete for Current Status: -

ボリュームの空き容量不足により転送に失敗していますね。

Storage Efficiency、ボリューム、aggregate、Snapshotを確認します。

::*> volume efficiency show -volume vol1, vol1_dst3 -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression
vserver volume state   policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume
------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- -----------------------------------------
svm     vol1   Enabled auto   false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst3
               Disabled
                       -      false       true               efficient               false         true            true                              true                          false
2 entries were displayed.

::*> volume show -volume vol1, vol1_dst3 -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 5.83GB    16GB            15.20GB 9.37GB 61%          825.3MB            8%                         825.3MB             10.17GB      67%                  -                 10.17GB             -                                   -
svm2    vol1_dst3
               9GB  16KB      9GB             9GB     9.00GB 99%          92.31MB            1%                         8.24GB              9.05GB       101%                 -                 9.05GB              0B                                  0%
2 entries were displayed.

::*> volume show-footprint -volume vol1_dst3


      Vserver : svm2
      Volume  : vol1_dst3

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            9.00GB       1%
             Footprint in Performance Tier             9.09GB     100%
             Footprint in FSxFabricpoolObjectStore         0B       0%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                             0B       0%
      Delayed Frees                                   89.38MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  9.09GB       1%

      Effective Total Footprint                        9.09GB       1%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 282.1GB
                               Total Physical Used: 29.97GB
                    Total Storage Efficiency Ratio: 9.41:1
Total Data Reduction Logical Used Without Snapshots: 53.97GB
Total Data Reduction Physical Used Without Snapshots: 24.30GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.22:1
Total Data Reduction Logical Used without snapshots and flexclones: 53.97GB
Total Data Reduction Physical Used without snapshots and flexclones: 24.30GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.22:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 207.8GB
Total Physical Used in FabricPool Performance Tier: 25.45GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 8.17:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 44.69GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 19.80GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.26:1
                Logical Space Used for All Volumes: 53.97GB
               Physical Space Used for All Volumes: 33.86GB
               Space Saved by Volume Deduplication: 20.11GB
Space Saved by Volume Deduplication and pattern detection: 20.11GB
                Volume Deduplication Savings ratio: 1.59:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.59:1
               Logical Space Used by the Aggregate: 38.92GB
              Physical Space Used by the Aggregate: 29.97GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.30:1
              Logical Size Used by Snapshot Copies: 228.1GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 30.98:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 30.98:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 1

::*> snapshot show -volume vol1, vol1_dst3
                                                                 ---Blocks---
Vserver  Volume   Snapshot                                  Size Total% Used%
-------- -------- ------------------------------------- -------- ------ -----
svm      vol1
                  test.2023-12-22_0533                     160KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                         24.45MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           312KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           212KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_041719
                                                           300KB     0%    0%
svm2     vol1_dst3
                  test.2023-12-22_0533                    1.29MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                           384KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           388KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           284KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           232KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           220KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           428KB     0%    0%
15 entries were displayed.

追加のSnapshotは作成されていませんでした。

転送先ボリュームでStorage Efficiencyの有効化

転送先ボリュームでStorage Efficiencyを有効化します。

::*> volume efficiency on -vserver svm2 -volume vol1_dst3

Error: command failed: Failed to enable efficiency on volume "vol1_dst3" of Vserver "svm2": A SnapMirror transfer is running or paused. Use the
       "snapmirror show" command to view the status of the transfer. Retry this command when the transfer is complete, or run "snapmirror abort -hard
       true" to abort the SnapMirror transfer and clear any checkpoint data.

先ほどStorage Efficiencyを有効化した際と同様のエラーですね。

::*> snapmirror abort -destination-path svm2:vol1_dst3

Error: command failed: No transfer to abort.

::*> snapmirror abort -destination-path svm2:vol1_dst3 -hard true
Operation is queued: snapmirror abort for the relationship with destination "svm2:vol1_dst3".

SnapMirror relationshipを確認します。

::*> snapmirror show -destination-path svm2:vol1_dst3                                  Source Path: svm:vol1
                               Source Cluster: -
                               Source Vserver: svm
                                Source Volume: vol1
                             Destination Path: svm2:vol1_dst3
                          Destination Cluster: -
                          Destination Vserver: svm2
                           Destination Volume: vol1_dst3
                            Relationship Type: XDP
                      Relationship Group Type: none
                             Managing Vserver: svm2
                          SnapMirror Schedule: -
                       SnapMirror Policy Type: async-mirror
                            SnapMirror Policy: MirrorAllSnapshots
                                  Tries Limit: -
                            Throttle (KB/sec): unlimited
              Consistency Group Item Mappings: -
           Current Transfer Throttle (KB/sec): -
                                 Mirror State: Snapmirrored
                          Relationship Status: Idle
                      File Restore File Count: -
                       File Restore File List: -
                            Transfer Snapshot: -
                            Snapshot Progress: -
                               Total Progress: -
                    Network Compression Ratio: -
                          Snapshot Checkpoint: -
                              Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                    Newest Snapshot Timestamp: 12/28 02:16:28
                            Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                  Exported Snapshot Timestamp: 12/28 02:16:28
                                      Healthy: false
                              Relationship ID: e94d0e31-a526-11ee-981e-bdd56ead09c8
                          Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8
                     Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8
                         Current Operation ID: -
                                Transfer Type: -
                               Transfer Error: -
                           Last Transfer Type: update
                          Last Transfer Error: Transfer failed. (Volume access error (No space left on device))
                    Last Transfer Error Codes: 6620144, 5898547, 6684700
                           Last Transfer Size: -
      Last Transfer Network Compression Ratio: -
                       Last Transfer Duration: -
                           Last Transfer From: svm:vol1
                  Last Transfer End Timestamp: 12/28 04:26:04
                             Unhealthy Reason: Transfer failed.
                        Progress Last Updated: -
                      Relationship Capability: 8.2 and above
                                     Lag Time: 2:38:8
                    Current Transfer Priority: -
                             SMTape Operation: -
                 Destination Volume Node Name: FsxId0ab6f9b00824a187c-01
                 Identity Preserve Vserver DR: -
                 Number of Successful Updates: 1
                     Number of Failed Updates: 1
                 Number of Successful Resyncs: 0
                     Number of Failed Resyncs: 0
                  Number of Successful Breaks: 0
                      Number of Failed Breaks: 0
                         Total Transfer Bytes: 9060359240
               Total Transfer Time in Seconds: 112
                Source Volume MSIDs Preserved: -
                                       OpMask: ffffffffffffffff
                       Is Auto Expand Enabled: -
          Percent Complete for Current Status: -

Snapshot Checkpointが265.8KBから-になっています。

Storage Efficiencyを有効化します。

::*> volume efficiency on -vserver svm2 -volume vol1_dst3
Efficiency for volume "vol1_dst3" of Vserver "svm2" is enabled.

::*>
::*> volume efficiency show -volume vol1, vol1_dst3 -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compressionvserver volume state   policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume
------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- -----------------------------------------
svm     vol1   Enabled auto   false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst3
               Enabled -      false       true               efficient               false         true            true                              true                          false
2 entries were displayed.

::*> volume show -volume vol1, vol1_dst3 -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 5.83GB    16GB            15.20GB 9.37GB 61%          825.3MB            8%                         825.3MB             10.17GB      67%                  -                 10.17GB             -                                   -
svm2    vol1_dst3
               9GB  693.5MB   9GB             9GB     8.32GB 92%          774.5MB            8%                         8.24GB              9.05GB       101%                 -                 9.04GB              0B                                  0%
2 entries were displayed.

::*> volume show-footprint -volume vol1_dst3


      Vserver : svm2
      Volume  : vol1_dst3

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            8.32GB       1%
             Footprint in Performance Tier             8.57GB     100%
             Footprint in FSxFabricpoolObjectStore         0B       0%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                        56.82MB       0%
      Delayed Frees                                   249.0MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  8.62GB       1%

      Effective Total Footprint                        8.62GB       1%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 282.1GB
                               Total Physical Used: 29.39GB
                    Total Storage Efficiency Ratio: 9.60:1
Total Data Reduction Logical Used Without Snapshots: 53.95GB
Total Data Reduction Physical Used Without Snapshots: 23.74GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.27:1
Total Data Reduction Logical Used without snapshots and flexclones: 53.95GB
Total Data Reduction Physical Used without snapshots and flexclones: 23.74GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.27:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 207.8GB
Total Physical Used in FabricPool Performance Tier: 24.87GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 8.35:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 44.68GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 19.25GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.32:1
                Logical Space Used for All Volumes: 53.95GB
               Physical Space Used for All Volumes: 33.17GB
               Space Saved by Volume Deduplication: 20.77GB
Space Saved by Volume Deduplication and pattern detection: 20.77GB
                Volume Deduplication Savings ratio: 1.63:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.63:1
               Logical Space Used by the Aggregate: 38.33GB
              Physical Space Used by the Aggregate: 29.39GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.30:1
              Logical Size Used by Snapshot Copies: 228.1GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 30.98:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 30.98:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 0

::*> snapshot show -volume vol1, vol1_dst3
                                                                 ---Blocks---
Vserver  Volume   Snapshot                                  Size Total% Used%
-------- -------- ------------------------------------- -------- ------ -----
svm      vol1
                  test.2023-12-22_0533                     160KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                         24.45MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           312KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           212KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_041719
                                                           300KB     0%    0%
svm2     vol1_dst3
                  test.2023-12-22_0533                    1.29MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                           384KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           388KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           284KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           232KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           220KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           152KB     0%    0%
15 entries were displayed.

ボリューム使用量が9.00GBから8.32GBになっています。チェックポイントを削除することで、途中まで転送された内容が破棄されたことが分かります。

SnapMirror update

それでは再度SnapMirror updateを行います。

::*> snapmirror update -destination-path svm2:vol1_dst3Operation is queued: snapmirror update of destination "svm2:vol1_dst3".

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updatedsource-path destination-path state        status       total-progress progress-last-updated----------- ---------------- ------------ ------------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Transferring 34.94MB        12/28 04:57:48

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status       total-progress progress-last-updated
----------- ---------------- ------------ ------------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Transferring 682.8MB        12/28 04:58:35

::*> snapmirror show -destination-path svm2:vol1_dst3 -fields state, status, total-progress, progress-last-updated
source-path destination-path state        status total-progress progress-last-updated
----------- ---------------- ------------ ------ -------------- ---------------------
svm:vol1    svm2:vol1_dst3   Snapmirrored Idle   -              -

::*> snapmirror show -destination-path svm2:vol1_dst3

                                  Source Path: svm:vol1
                               Source Cluster: -
                               Source Vserver: svm
                                Source Volume: vol1
                             Destination Path: svm2:vol1_dst3
                          Destination Cluster: -
                          Destination Vserver: svm2
                           Destination Volume: vol1_dst3
                            Relationship Type: XDP
                      Relationship Group Type: none
                             Managing Vserver: svm2
                          SnapMirror Schedule: -
                       SnapMirror Policy Type: async-mirror
                            SnapMirror Policy: MirrorAllSnapshots
                                  Tries Limit: -
                            Throttle (KB/sec): unlimited
              Consistency Group Item Mappings: -
           Current Transfer Throttle (KB/sec): -
                                 Mirror State: Snapmirrored
                          Relationship Status: Idle
                      File Restore File Count: -
                       File Restore File List: -
                            Transfer Snapshot: -
                            Snapshot Progress: -
                               Total Progress: -
                    Network Compression Ratio: -
                          Snapshot Checkpoint: 262.5KB
                              Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                    Newest Snapshot Timestamp: 12/28 02:16:28
                            Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                  Exported Snapshot Timestamp: 12/28 02:16:28
                                      Healthy: false
                              Relationship ID: e94d0e31-a526-11ee-981e-bdd56ead09c8
                          Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8
                     Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8
                         Current Operation ID: -
                                Transfer Type: -
                               Transfer Error: -
                           Last Transfer Type: update
                          Last Transfer Error: Transfer failed. (Volume access error (No space left on device))
                    Last Transfer Error Codes: 6620144, 5898547, 6684700
                           Last Transfer Size: -
      Last Transfer Network Compression Ratio: -
                       Last Transfer Duration: -
                           Last Transfer From: svm:vol1
                  Last Transfer End Timestamp: 12/28 05:06:22
                             Unhealthy Reason: Transfer failed.
                        Progress Last Updated: -
                      Relationship Capability: 8.2 and above
                                     Lag Time: 2:55:12
                    Current Transfer Priority: -
                             SMTape Operation: -
                 Destination Volume Node Name: FsxId0ab6f9b00824a187c-01
                 Identity Preserve Vserver DR: -
                 Number of Successful Updates: 1
                     Number of Failed Updates: 2
                 Number of Successful Resyncs: 0
                     Number of Failed Resyncs: 0
                  Number of Successful Breaks: 0
                      Number of Failed Breaks: 0
                         Total Transfer Bytes: 9060359240
               Total Transfer Time in Seconds: 112
                Source Volume MSIDs Preserved: -
                                       OpMask: ffffffffffffffff
                       Is Auto Expand Enabled: -
          Percent Complete for Current Status: -

ボリュームの空き容量不足で転送が中断されました。

Storage Efficiencyの実行状態を確認します。

::*> volume efficiency show -volume vol1_dst3
Vserver    Volume           State     Status      Progress           Policy
---------- ---------------- --------- ----------- ------------------ ----------
svm2       vol1_dst3        Enabled   Idle        Idle for 00:03:10  -

特に実行はされていません。

Storage Efficiency、ボリューム、aggregate、Snapshotを確認します。

::*> volume efficiency show -volume vol1, vol1_dst3 -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression
vserver volume state   policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume
------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- -----------------------------------------
svm     vol1   Enabled auto   false       true               efficient               false         true            true                              true                          false
svm2    vol1_dst3
               Enabled -      false       true               efficient               false         true            true                              true                          false
2 entries were displayed.

::*> volume show -volume vol1, vol1_dst3 -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available
vserver volume size available filesystem-size total   used   percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent
------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- -------------------------------------------
svm     vol1   16GB 5.83GB    16GB            15.20GB 9.37GB 61%          825.3MB            8%                         825.3MB             10.17GB      67%                  -                 10.17GB             -                                   -
svm2    vol1_dst3
               9GB  72KB      9GB             9GB     9.00GB 99%          95.56MB            1%                         8.24GB              9.06GB       101%                 -                 9.05GB              0B                                  0%
2 entries were displayed.

::*> volume show-footprint -volume vol1_dst3


      Vserver : svm2
      Volume  : vol1_dst3

      Feature                                         Used       Used%
      --------------------------------             ----------    -----
      Volume Data Footprint                            9.00GB       1%
             Footprint in Performance Tier             9.01GB     100%
             Footprint in FSxFabricpoolObjectStore         0B       0%
      Volume Guarantee                                     0B       0%
      Flexible Volume Metadata                        50.61MB       0%
      Delayed Frees                                    6.29MB       0%
      File Operation Metadata                             4KB       0%

      Total Footprint                                  9.05GB       1%

      Effective Total Footprint                        9.05GB       1%

::*> aggr show-efficiency -instance

                             Name of the Aggregate: aggr1
                      Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01
Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 282.1GB
                               Total Physical Used: 30.04GB
                    Total Storage Efficiency Ratio: 9.39:1
Total Data Reduction Logical Used Without Snapshots: 53.96GB
Total Data Reduction Physical Used Without Snapshots: 24.36GB
Total Data Reduction Efficiency Ratio Without Snapshots: 2.21:1
Total Data Reduction Logical Used without snapshots and flexclones: 53.96GB
Total Data Reduction Physical Used without snapshots and flexclones: 24.36GB
Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.21:1
Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 207.8GB
Total Physical Used in FabricPool Performance Tier: 25.52GB
Total FabricPool Performance Tier Storage Efficiency Ratio: 8.14:1
Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 44.69GB
Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 19.87GB
Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.25:1
                Logical Space Used for All Volumes: 53.96GB
               Physical Space Used for All Volumes: 33.85GB
               Space Saved by Volume Deduplication: 20.11GB
Space Saved by Volume Deduplication and pattern detection: 20.11GB
                Volume Deduplication Savings ratio: 1.59:1
                 Space Saved by Volume Compression: 0B
                  Volume Compression Savings ratio: 1.00:1
      Space Saved by Inline Zero Pattern Detection: 0B
                    Volume Data Reduction SE Ratio: 1.59:1
               Logical Space Used by the Aggregate: 38.98GB
              Physical Space Used by the Aggregate: 30.04GB
           Space Saved by Aggregate Data Reduction: 8.94GB
                 Aggregate Data Reduction SE Ratio: 1.30:1
              Logical Size Used by Snapshot Copies: 228.1GB
             Physical Size Used by Snapshot Copies: 7.36GB
              Snapshot Volume Data Reduction Ratio: 30.98:1
            Logical Size Used by FlexClone Volumes: 0B
          Physical Sized Used by FlexClone Volumes: 0B
             FlexClone Volume Data Reduction Ratio: 1.00:1
Snapshot And FlexClone Volume Data Reduction SE Ratio: 30.98:1
                         Number of Volumes Offline: 0
                    Number of SIS Disabled Volumes: 3
         Number of SIS Change Log Disabled Volumes: 0

::*> snapshot show -volume vol1, vol1_dst3
                                                                 ---Blocks---
Vserver  Volume   Snapshot                                  Size Total% Used%
-------- -------- ------------------------------------- -------- ------ -----
svm      vol1
                  test.2023-12-22_0533                     160KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                         24.45MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           312KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           148KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           292KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_045744
                                                           156KB     0%    0%
svm2     vol1_dst3
                  test.2023-12-22_0533                    1.29MB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
                                                           384KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
                                                           388KB     0%    0%
                  snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507
                                                           284KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406
                                                           232KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027
                                                           220KB     0%    0%
                  snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779294.2023-12-28_021628
                                                           432KB     0%    0%
15 entries were displayed.

SnapMirror転送中にStorage Efficiencyによりデータが削減され、転送しきれるかと思ったのですができませんでした。

転送先のボリュームサイズは転送元のボリュームサイズ以上のものを準備しよう

SnapMirrorの転送先ボリュームのサイズ不足である場合の挙動を確認してみました。

冒頭で紹介した方式は実現できないことが確認できました。カスケードSnapMirrorを活用する方法でStorage Efficiencyをかけたデータを最終的な転送先FSxNファイルシステムに転送するのが良いでしょう。

転送先のボリュームサイズは転送元のボリュームサイズ以上のものを準備しましょう。なお、ボリュームタイプがDPのボリュームはデフォルトでボリュームの自動拡張が有効です。

また、チェックポイントを活用することで、途中まで転送されている情報を引き続き転送できることも確認できました。

移行中にSnapMirrorベースライン転送が失敗した場合、ネットワーク接続の切断、転送の中止、コントローラ フェイルオーバーなど、様々な原因があります。失敗の原因を修正後、再開チェックポイントがある場合はSnapMirror転送を再開できます。

失敗したSnapMirrorベースライン転送の再開

この記事が誰かの助けになれば幸いです。

以上、AWS事業本部 コンサルティング部の のんピ(@non____97)でした!

Share this article

facebook logohatena logotwitter logo

© Classmethod, Inc. All rights reserved.